In speech recognition, it is essential to model the phonetic content of the input signal while discarding irrelevant factors such as speaker variations and noise, which is challenging in low-resource settings. Self-supervised pre-training has been proposed as a way to improve both supervised and unsupervised speech recognition, including frame-level feature representations and Acoustic Word Embeddings (AWE) for variable-length segments. However, self-supervised models alone cannot learn perfect separation of the linguistic content as they are trained to optimize indirect objectives. In this work, we experiment with different pre-trained self-supervised features as input to AWE models and show that they work best within a supervised framework. Models trained on English can be transferred to other languages with no adaptation and outperform self-supervised models trained solely on the target languages.
translated by 谷歌翻译
一种共同的销售策略涉及让帐户高管(AES)积极联系并与潜在客户联系。但是,并非所有的接触尝试都有积极的效果:有些尝试不会改变客户的决策,而另一些尝试甚至可能会干扰所需的结果。在这项工作中,我们建议使用因果推断来估计与每个潜在客户联系并相应地制定联系政策的效果。我们从在线珠宝市场worthy.com上证明了这种方法。我们研究了有价值的业务流程,以确定相关的决策和结果,并对他们制定的方式进行正式的假设。使用因果工具,我们选择了一个决策点,改善AE接触活动似乎是有希望的。然后,我们制定了一个个性化的政策,建议仅与对其有益的客户联系。最后,我们在3个月内验证了A \ B测试中的结果,从而导致目标人群的项目交付率增加了22%(p值= 0.026)。现在,该政策正在持续使用。
translated by 谷歌翻译
数据科学有可能改善各种垂直领域的业务。尽管狮子的数据科学项目使用了一种预测方法,但这些预测应成为决策。但是,这种两步的方法不仅是最佳的,甚至可能降低性能并使项目失败。另一种选择是遵循规范性的框架,在该框架中,行动是“第一公民”,以便该模型制定规定采取行动的政策,而不是预测结果。在本文中,我们解释了为什么规定的方法很重要,并提供了分步方法论:规定的画布。后者旨在改善项目利益相关者的框架和沟通,包括项目和数据科学经理,以成功地产生业务影响。
translated by 谷歌翻译
人机交互的参与是参与互动的实体建立,维护和结束他们感知连接的过程。必须监测各种基于AI的医疗范式范式的患者的参与状态。这包括改变社会行为的医疗条件,例如自闭症谱系障碍(ASD)或注意力缺陷/多动障碍(ADHD)。订婚是一种多方面构造,由行为,情感和精神组成部分组成。以前的研究忽视了参与的多面条性质。在本文中,提出了一种系统以使用上下文和关系特征来区分这些方面。这可以促进进一步细粒度的分析。将多种机器学习分类器包括传统和深度学习模型,以获得此任务。在具有基于神经网络的分类的22242个实例的平衡数据集上,可以获得具有F分数和0.74和0.23的F分和0.23%的最高精度。
translated by 谷歌翻译
阳性是来自观察数据的因果推断的三种条件之一。验证积极性的标准方法是分析倾向的分布。然而,为了民主化非专家做因果推断的能力,需要设计一种算法(i)测试阳性和(ii)解释在缺乏协变量阳性的地方。后者可用于建议违反阳性的进一步因果分析和/或鼓励实验的限制。本文的贡献首先存在自动正性分析的问题,其次是基于两个步骤过程提出算法。第一步,模拟协变量上的倾向条件,然后使用多个假设检测分析后一分布,以产生积极违规标签。第二步使用非对称修剪的决策树以解释。后者进一步转换为可读文本,非专家可以理解。我们在大型软件企业的专有数据集上展示了我们的方法。
translated by 谷歌翻译
The use of supervised Machine Learning (ML) to enhance Intrusion Detection Systems has been the subject of significant research. Supervised ML is based upon learning by example, demanding significant volumes of representative instances for effective training and the need to re-train the model for every unseen cyber-attack class. However, retraining the models in-situ renders the network susceptible to attacks owing to the time-window required to acquire a sufficient volume of data. Although anomaly detection systems provide a coarse-grained defence against unseen attacks, these approaches are significantly less accurate and suffer from high false-positive rates. Here, a complementary approach referred to as 'One-Shot Learning', whereby a limited number of examples of a new attack-class is used to identify a new attack-class (out of many) is detailed. The model grants a new cyber-attack classification without retraining. A Siamese Network is trained to differentiate between classes based on pairs similarities, rather than features, allowing to identify new and previously unseen attacks. The performance of a pre-trained model to classify attack-classes based only on one example is evaluated using three datasets. Results confirm the adaptability of the model in classifying unseen attacks and the trade-off between performance and the need for distinctive class representation.
translated by 谷歌翻译
The success of CNNs in various applications is accompanied by a significant increase in the computation and parameter storage costs. Recent efforts toward reducing these overheads involve pruning and compressing the weights of various layers without hurting original accuracy. However, magnitude-based pruning of weights reduces a significant number of parameters from the fully connected layers and may not adequately reduce the computation costs in the convolutional layers due to irregular sparsity in the pruned networks. We present an acceleration method for CNNs, where we prune filters from CNNs that are identified as having a small effect on the output accuracy. By removing whole filters in the network together with their connecting feature maps, the computation costs are reduced significantly. In contrast to pruning weights, this approach does not result in sparse connectivity patterns. Hence, it does not need the support of sparse convolution libraries and can work with existing efficient BLAS libraries for dense matrix multiplications. We show that even simple filter pruning techniques can reduce inference costs for VGG-16 by up to 34% and ResNet-110 by up to 38% on CIFAR10 while regaining close to the original accuracy by retraining the networks.
translated by 谷歌翻译